Newest journal articles by subject

Subjects > Computer Science

Gemma 3 Technical Report
Gemma Team Aishwarya Kamath;Johan FerretShreya PathakNino VieillardRamona MerhejSarah PerrinTatiana MatejovicovaAlexandre Ram'eMorgane RivièreLouis RouillardThomas MesnardGeoffrey CideronJean-Bastien GrillSabela RamosEdouard YvinecMichelle CasbonEtienne PotIvo PenchevGael LiuFrancesco VisinKathleen KenealyLucas BeyerXiaohai ZhaiAnton TsitsulinR. Busa-FeketeAlex FengNoveen SachdevaBenjamin ColemanYi GaoBasil MustafaIain BarrEmilio ParisottoDavid TianMatan EyalColin CherryJan-Thorsten PeterDanila SinopalnikovSurya BhupatirajuRishabh AgarwalMehran KazemiDan MalkinRavin KumarDavid VilarIdan BrusilovskyJiaming LuoA. SteinerAbe FriesenAbhanshu SharmaAbheesht SharmaAdi Mayrav GiladyAdrian GoedeckemeyerAlaa SaadeAlexander KolesnikovAlexei BendeburyAlvin AbdagicAmit VadiAndr'as GyorgyAndré Susano PintoAnil DasAnkur BapnaAntoine MiechAntoine YangAntonia PatersonAshish ShenoyAyan ChakrabartiBilal PiotBoxi WuBobak ShahriariBryce PetriniCharlie ChenCharline Le LanChristopher A. Choquette-ChooCj CareyC. BrickDaniel DeutschDanielle EisenbudDee CattleDerek ChengDimitris PaparasDivyashree Shivakumar SreepathihalliDoug ReidDustin TranDustin ZelleEric NolandErwin HuizengaE. KharitonovFrederick LiuG. AmirkhanyanGlenn CameronHadi HashemiHanna Klimczak-Pluci'nskaHarman SinghHarsh MehtaHarshal Tushar LehriHussein HazimehIan BallantyneIdan SzpektorIvan NardiniJean Pouget-AbadieJetha ChanJoe StantonJ. Michael WietingJ. LaiJordi OrbayJoe FernandezJoshua NewlanJunsong JiJyotinder SinghKat BlackKathy YuKevin HuiKiran VodrahalliKlaus GreffLinhai QiuMarcella ValentineMarina CoelhoMarvin RitterMatt HoffmanMatthew WatsonMayank ChaturvediMichael MoynihanMin MaNabila BabarNatasha NoyNathan ByrdNick RoyNikola MomchevNilay ChauhanOskar BunyanPankil BotardaPaul CaronP. RubensteinPhil CullitonP. SchmidPier Giuseppe SessaPing-mei XuP. StańczykP. TaftiRakesh ShivannaRenjie WuRenke PanR. RokniRob WilloughbyRohith ValluRyan MullinsSammy JeromeSara SmootSertan GirginShariq IqbalShashir ReddyShruti ShethSiim PõderSijal BhatnagarSindhu Raghuram PanyamSivan EigerSusan ZhangTianqi LiuTrevor YacovoneT. LiechtyUday KalraUtku EvciVedant MisraVincent RoseberryVladimir FeinbergVlad KolesnikovWoohyun HanWoosuk KwonXi ChenYinlam ChowYuvein ZhuZichuan WeiZ. EgyedVictor CotrutaMinh GiangPhoebe KirkAnand RaoJessica LoErica MoreiraLuiz Gustavo MartinsOmar SansevieroLucas GonzalezZach GleicherTris WarkentinV. MirrokniEvan SenterEli CollinsJoelle BarralZ. GhahramaniR. HadsellYossi MatiasD. SculleySlav PetrovNoah FiedelNoam ShazeerO. VinyalsJeffrey DeanD. HassabisK. KavukcuogluClément FarabetElena BuchatskayaJean-Baptiste AlayracRohan AnilDmitry LepikhinSebastian BorgeaudOlivier BachemArmand JoulinAlek AndreevCassidy HardinRobert DadashiL'eonard Hussenot
ArXiv Published 2025/03/25

Summary:

A novel post-training recipe significantly improves the math, chat, instruction-following and multilingual abilities, making Gemma3-4B-IT competitive with Gemma2-27B-IT and Gemma3-27B-IT comparable to Gemini-1.5-Pro across benchmarks.

FLUX.1 Kontext: Flow Matching for In-Context Image Generation and Editing in Latent Space
Black Forest Labs;Stephen BatifolA. BlattmannFrederic BoeselSaksham ConsulCyril DiagneTim DockhornJack EnglishZion EnglishPatrick EsserSumith KulalKyle LaceyYam LeviCheng LiDominik LorenzJonas MullerDustin PodellRobin RombachHarry SainiAxel SauerLuke Smith
ArXiv Published 2025/06/17

Summary:

Compared to current editing models that exhibit degradation in character consistency and stability across multiple turns, it is observed that FLUX.1 Kontext improved preservation of objects and characters, leading to greater robustness in iterative workflows.

SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features
Michael Tschannen;Alexey GritsenkoXiao WangM. NaeemIbrahim M. AlabdulmohsinNikhil ParthasarathyTalfan EvansLucas BeyerYe XiaBasil MustafaOlivier H'enaffJeremiah HarmsenA. SteinerXiao-Qi Zhai
ArXiv Published 2025/02/20

Summary:

This second iteration of SigLIP 2 introduces SigLIP 2, a family of new multilingual vision-language encoders that build on the success of the original SigLIP, and extends the original image-text training objective with several prior, independently developed techniques into a unified recipe.

GR00T N1: An Open Foundation Model for Generalist Humanoid Robots
Nvidia;Johan BjorckFernando CastañedaNikita CherniadevXingye DaRunyu DingLinxiJimFanYu FangDieter FoxFengyuan HuSpencer HuangJ. JangZhenyuan JiangJan KautzKaushil KundaliaLawrence LaoZhiqi LiZongyu LinKevin LinGuilin LiuEdith LlontopLoic MagneA. MandlekarAvnish NarayanSoroush NasirianyScott ReedY. TanGuanzhi WangZu WangJing WangQi WangJiannan XiangYuqi XieYinzhen XuZhen-Teng XuSeonghyeon YeZhiding YuAo ZhangHao ZhangYizhou ZhaoRuijie ZhengYuke Zhu
ArXiv Published 2025/03/18

Summary:

This work introduces GR00T N1, an open foundation model for humanoid robots that outperforms the state-of-the-art imitation learning baselines on standard simulation benchmarks across multiple robot embodiments and deploys the model on the Fourier GR-1 humanoid robot for language-conditioned bimanual manipulation tasks.

Toward expert-level medical question answering with large language models
Karan Singhal;Tao TuJuraj GottweisR. SayresEllery WulczynMohamed AminLe HouKevin ClarkStephen R. PfohlHeather Cole-LewisDarlene NealQ. RashidMike SchaekermannAmy WangD. DashJonathan H. ChenNigam H. ShahSami LachgarP. MansfieldSushant PrakashBradley GreenEwa DominowskaBlaise Agüera y ArcasNenad TomaševYun LiuRenee WongChristopher SemtursS. MahdaviJoelle K. BarralDale R. WebsterG. CorradoYossi MatiasShekoofeh AziziA. KarthikesalingamVivek Natarajan
Nature Medicine Published 2025/01/08

Summary:

With an improved framework for model development and evaluation, a large language model is shown to provide answers to medical questions that are comparable or preferred with respect to those provided by human physicians.

Transparency In The reporting of Artificial INtelligence – the TITAN guideline
R. Agha;Ginimol MathewRasha RashidAhmed KerwanA. Al-JabirC. SohrabiT. FranchiMaria NicolaM. Agha
Premier Journal of Science

Summary:

A guideline to transparently reporting the use of AI in any manuscript in general is presented and will evolve over time as technology, systems and behaviour evolve.

Scaling Rectified Flow Transformers for High-Resolution Image Synthesis
Patrick Esser;Sumith KulalA. BlattmannRahim EntezariJonas MullerHarry SainiYam LeviDominik LorenzAxel SauerFrederic BoeselDustin PodellTim DockhornZion EnglishKyle LaceyAlex GoodwinYannik MarekRobin Rombach
Published 2024/03/05

Summary:

This work improves existing noise sampling techniques for training rectified flow models by biasing them towards perceptually relevant scales and presents a novel transformer-based architecture for text-to-image generation that uses separate weights for the two modalities and enables a bidirectional flow of information between image and text tokens.

Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Marah Abdin;Sam Ade JacobsA. A. AwanJ. AnejaAhmed AwadallahH. AwadallaNguyen BachAmit BahreeArash BakhtiariHarkirat Singh BehlAlon BenhaimMisha BilenkoJohan BjorckSébastien BubeckMartin CaiC. C. T. MendesWeizhu ChenVishrav ChaudharyParul ChopraA. GiornoGustavo de RosaMatthew DixonRonen EldanVictor FragosoDan IterAbhishek GoswamiS. GunasekarEmman HaiderJunheng HaoRussell J. HewettJamie HuynhMojan JavaheripiXin JinPiero KauffmannNikos KarampatziakisDongwoo KimYoung Jin KimMahoud KhademiLev KurilenkoJames R. LeeYin Tat LeeYuanzhi LiChen LiangWeishung LiuEric LinZeqi LinPiyush MadanArindam MitraHardik ModiA. NguyenBrandon NorickBarun PatraD. Perez-BeckerThomas PortetReid PryzantHeyang QinMarko RadmilacLiliang RenCorby RossetSambudha RoyOlli SaarikiviA. SaiedAdil SalimMichael SantacroceShital ShahNing ShangHiteshi SharmaXianmin SongOlatunji RuwasePraneetha VaddamanuXin WangRachel WardGuanhua WangP. WitteMichael WyattCan XuJiahang XuSonal YadavFan YangZiyi YangDonghan YuCheng-yuan ZhangCyril ZhangJianwen ZhangL. ZhangYi ZhangYunan ZhangXiren ZhouYifan Yang
ArXiv Published 2024/04/22
Accurate structure prediction of biomolecular interactions with AlphaFold 3
Josh Abramson;Jonas AdlerJack DungerRichard EvansTim GreenA. PritzelOlaf RonnebergerLindsay WillmoreAndrew J BallardJoshua BambrickSebastian BodensteinDavid A EvansChia-Chun HungMichael O’NeillD. ReimanKathryn TunyasuvunakoolZachary WuAkvile ZemgulyteEirini ArvanitiCharles BeattieOttavia BertolliAlex BridglandAlexey CherepanovMiles CongreveA. Cowen-RiversAndrew CowieMichael FigurnovFabian B FuchsHannah GladmanRishub JainYousuf A. KhanCaroline M R LowKuba PerlinAnna PotapenkoPascal SavySukhdeep SinghA. SteculaAshok ThillaisundaramCatherine TongSergei YakneenEllen D. ZhongMichal ZielinskiAugustin ŽídekV. BapstPushmeet KohliMax JaderbergD. HassabisJ. Jumper
Nature Published 2024/05/08

Summary:

The new AlphaFold model demonstrates substantially improved accuracy over many previous specialized tools: far greater accuracy for protein–ligand interactions compared with state-of-the-art docking tools, much higher accuracy for protein–nucleic acid interactions compared with nucleic-acid-specific predictors and substantially higher antibody–antigen prediction accuracy.

Gemma 2: Improving Open Language Models at a Practical Size
Gemma Team Morgane Riviere;Shreya PathakPier Giuseppe SessaCassidy HardinSurya BhupatirajuL'eonard HussenotThomas MesnardBobak ShahriariAlexandre Ram'eJohan FerretPeter LiuP. TaftiAbe FriesenMichelle CasbonSabela RamosRavin KumarCharline Le LanSammy JeromeAnton TsitsulinNino VieillardP. StańczykSertan GirginNikola MomchevMatt HoffmanS. ThakoorJean-Bastien GrillBehnam NeyshaburAlanna WaltonA. SeverynAlicia ParrishAliya AhmadAllen HutchisonAlvin AbdagicAmanda CarlAmy ShenAndy BrockAndy CoenenAnthony LaforgeAntonia PatersonBen BastianBilal PiotBoxi WuBrandon RoyalCharlie ChenChintu KumarChris PerryChristoper A. WeltyChristopher A. Choquette-ChooDanila SinopalnikovDavid WeinbergerDimple VijaykumarDominika Rogozi'nskaD. HerbisonElisa BandyEmma WangEric NolandErica MoreiraEvan SenterEvgenii EltyshevFrancesco VisinGabriel RasskinGary WeiGlenn CameronGus MartinsHadi HashemiHanna Klimczak-Pluci'nskaHarleen BatraH. DhandIvan NardiniJacinda MeinJack ZhouJames SvenssonJ. StanwayJetha ChanJin ZhouJoana CarrasqueiraJoana IljaziJocelyn BeckerJoe FernandezJoost R. van AmersfoortJosh GordonJosh LipschultzJoshua NewlanJunsong JiKareem MohamedKartikeya BadolaKat BlackKatie MillicanK. McDonellK. NguyenKiranbir SodhiaKish GreeneLars Lowe SjoesundLauren UsuiL. SifreLena HeuermannLeti-cia LagoLilly McNealusLivio Baldini SoaresLogan KilpatrickLucas DixonL. MartinsMachel ReidManvinder SinghMark IversonMartin GornerMat VellosoMateo WirthMatt DavidowMatt MillerMatthew RahtzMatthew WatsonMeg RisdalMehran KazemiMichael MoynihanMing ZhangMinsuk KahngMinwoo ParkMofi RahmanM. KhatwaniNatalie DaoNen-shad BardoliwallaN. DevanathanNeta DumaiNilay ChauhanOscar WahltinezPankil BotardaParker BarnesP. BarhamPaul MichelPeng-chong JinPetko GeorgievPhil CullitonPradeep KuppalaR. ComanescuRamona MerhejReena JanaR. RokniRishabh AgarwalRyan MullinsSamaneh SaadatS. M. CarthySarah PerrinSébastien M. R. ArnoldSe-bastian KrauseShengyang DaiS. GargShruti ShethS. RonstromSusan ChanTimothy JordanTing YuTom EcclesT. HenniganTomás KociskýTulsee DoshiVihan JainVikas YadavVilobh MeshramVishal DharmadhikariWarren BarkleyWei WeiWenming YeWoohyun HanWoosuk KwonXiang XuZhe ShenZhitao GongZichuan WeiVictor CotrutaPhoebe KirkAnand RaoMinh GiangLudovic PeranTris WarkentinEli CollinsJoelle BarralZ. GhahramaniR. HadsellD. SculleyJeanine BanksAnca DraganSlav PetrovO. VinyalsJeffrey DeanD. HassabisK. KavukcuogluClément FarabetElena BuchatskayaSebastian BorgeaudNoah FiedelArmand JoulinKathleen KenealyRobert DadashiAlek Andreev
ArXiv Published 2024/07/31

Summary:

Gemma 2, a new addition to the Gemma family of lightweight, state-of-the-art open models, ranging in scale from 2 billion to 27 billion parameters, delivers the best performance for their size, and even offers competitive alternatives to models that are 2-3 times bigger.

OpenVLA: An Open-Source Vision-Language-Action Model
Moo Jin Kim;Karl PertschSiddharth KaramchetiTed XiaoA. BalakrishnaSuraj NairRafael RafailovE. FosterGrace LamPannag R. SanketiQuan VuongThomas KollarBenjamin BurchfielRuss TedrakeDorsa SadighSergey LevinePercy LiangChelsea Finn
ArXiv Published 2024/06/13

Summary:

OpenVLA, a 7B-parameter open-source VLA trained on a diverse collection of 970k real-world robot demonstrations, is introduced and it is shown that it can effectively fine-tune OpenVLA for new settings, with especially strong generalization results in multi-task environments involving multiple objects and strong language grounding abilities.

Gemma: Open Models Based on Gemini Research and Technology
Gemma Team Thomas Mesnard;Cassidy HardinRobert DadashiSurya BhupatirajuShreya PathakL. SifreMorgane RivièreMihir KaleJ Christopher LoveP. TaftiL'eonard HussenotA. ChowdheryAdam RobertsAditya BaruaAlex BotevAlex Castro-RosAmbrose SloneAm'elie H'eliouAndrea TacchettiAnna BulanovaAntonia PatersonBeth TsaiBobak ShahriariCharline Le LanChristopher A. Choquette-ChooClé-ment CrepyDaniel CerDaphne IppolitoDavid ReidElena BuchatskayaEric NiEric NolandGeng YanGeorge TuckerGeorge-Christian MuraruGrigory RozhdestvenskiyH. MichalewskiIan TenneyIvan GrishchenkoJacob AustinJames KeelingJane LabanowskiJean-Baptiste LespiauJ. StanwayJenny BrennanJeremy ChenJohan FerretJustin ChiuJ. Mao-JonesKather-ine LeeKathy YuKatie MillicanLars Lowe SjoesundLisa LeeLucas DixonMachel ReidMaciej MikułaMateo WirthMichael SharmanNikolai ChinaevNithum ThainOlivier BachemOs-car ChangOscar WahltinezPaige BaileyPaul MichelPetko YotovPier Giuseppe SessaR. ChaabouniR. ComanescuReena JanaRohan AnilRoss McilroyRuibo LiuRyan MullinsSamuel L. SmithSebastian BorgeaudSertan GirginSholto DouglasShree PandyaSiamak ShakeriSoham DeTed KlimenkoT. HenniganVladimir FeinbergWojciech StokowiecYu-Hui ChenZafarali AhmedZhitao GongTris WarkentinLudovic PeranMinh GiangClément FarabetO. VinyalsJeffrey DeanK. KavukcuogluD. HassabisZ. GhahramaniDouglas EckJoelle BarralFernando PereiraEli CollinsArmand JoulinNoah FiedelEvan SenterAlek AndreevKathleen Kenealy
ArXiv Published 2024/03/13

Summary:

This work introduces Gemma, a family of lightweight, state-of-the art open models built from the research and technology used to create Gemini models, and presents comprehensive evaluations of safety and responsibility aspects of the models, alongside a detailed description of model development.

PyTorch 2: Faster Machine Learning Through Dynamic Python Bytecode Transformation and Graph Compilation
Jason Ansel;Edward YangHorace HeN. GimelsheinAnimesh JainMichael VoznesenskyBin BaoPeter BellD. BerardEvgeni BurovskiGeeta ChauhanAnjali ChourdiaW. ConstableAlban DesmaisonZachary DeVitoElias EllisonW. FengJiong GongMichael GschwindB. HirshSherlock HuangKshiteej KalambarkarLaurent KirschMichael LazosM. LezcanoYanbo LiangJason LiangYinghai LuC. LukBertrand A. MaherYunjie PanChristian PuhrschMatthias ResoMark-Albert SaroufimMarcos Yukio SiraichiHelen SukShunting ZhangMichael SuoP. TilletXu ZhaoEikan WangKeren ZhouRichard ZouXiaodong WangAjit MathewsW. WenGregory ChananPeng WuSoumith Chintala
Proceedings of the 29th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 2 Published 2024/04/27

Summary:

Results show that TorchDynamo is able to capture graphs more robustly than prior approaches while adding minimal overhead, and TorchInductor is able to provide a 2.41× training geometric mean speedup on an NVIDIA A100 GPU across 180+ real-world models, which outperforms six other compilers.

The EMBL-EBI Job Dispatcher sequence analysis tools framework in 2024
F. Madeira;Nandana MadhusoodananJoon LeeAlberto EusebiAnia NiewielskaA. TiveyRodrigo LopezSarah Butcher
Nucleic Acids Research Published 2024/04/10

Summary:

Recent improvements to Job Dispatcher are overviews, including its brand new website and documentation, enhanced visualisations, improved job management, and a rising trend of user reliance on the service from low- and middle-income regions.

Cosmos World Foundation Model Platform for Physical AI
Nvidia Niket Agarwal;Arslan AliMaciej BalaYogesh BalajiErik BarkerTiffany CaiPrithvijit ChattopadhyayYongxin ChenYin CuiYifan DingDaniel DworakowskiJiaoJiao FanMichele FenziFrancesco FerroniSanja FidlerDieter FoxSongwei GeYunhao GeJinwei GuSiddharth GururaniEthan HeJiahui HuangJ. HuffmanPooya JannatyJingyi JinS. KimGergely Kl'arGrace LamShiyi LanL. Leal-TaixéAnqi LiZhaoshuo LiChen-Hsuan LinTsung-Yi LinHuan LingMing-Yu LiuXian LiuAlice LuoQianli MaHanzi MaoKaichun MoA. MousavianSeungjun NahSriharsha NivertyDavid PageDespoina PaschalidouZeeshan PatelLindsey PavaoMorteza RamezanaliF. RedaXiao-Shuai RenVasanth Rao Naik SabavatEd SchmerlingStella ShiBartosz StefaniakShitao TangLyne P. TchapmiPrzemek TredakWei-Cheng TsengJ. VargheseHao WangHaoxiang WangHengyi WangTingwei WangFangyin WeiXinyue WeiJay Zhangjie WuJiashu XuWei YangLin Yen-ChenXiaohui ZengYuan ZengJing ZhangQinsheng ZhangYuxuan ZhangQingqing ZhaoArtur Zolkowski
ArXiv Published 2025/01/06

Summary:

The Cosmos World Foundation Model Platform is presented to help developers build customized world models for their Physical AI setups and position a world foundation model as a general-purpose world model that can be fine-tuned into customized world models for downstream applications.

UniProt: the Universal Protein Knowledgebase in 2025
Alex Maria-Jesus Sandra Michele Aduragbemi Shadab Emily Bateman Martin Orchard Magrane Adesina Ahmad Bowle;Alex BatemanM. MartinSandra OrchardM. MagraneA. AdesinaShadab AhmadE. Bowler-BarnettHema Bye-A-JeeD. CarpentierPaulus DennyJun FanPenelope GarmiriLeonardo Jose da Costa GonzalesAbdulrahman HusseinAlexandr IgnatchenkoGiuseppe InsanaRizwan IshtiaqVishal JoshiDushyanth JyothiSwaathi KandasaamyA. LockAurélien LucianiJie LuoYvonne LussiJ. MarinPedro RaposoDan RiceRafael SantosElena SperettaJames L. StephensonPrabhat TotooNidhi TyagiNadya UrakovaPreethi VasudevKate WarnerSupun WijerathneConny Wing-Heng YuR. ZaruAlan J. BridgeL. AimoGhislaine Argoud-PuyA. AuchinclossK. AxelsenParit BansalDelphine BaratinTeresa M Batista NetoMarie-Claude BlatterJerven T. BollemanE. BoutetLionel BreuzaBlanca Cabrera GilCristina Casals-CasasKamal Chikh EchioukhE. CoudertB. CucheEdouard de CastroA. EstreicherM. FamigliettiM. FeuermannElisabeth GasteigerPascale GaudetS. GehantV. GerritsenA. GosNadine GruazC. HuloNevila Hyka-NouspikelF. JungoArnaud KerhornouP. MercierD. LieberherrP. MassonA. MorgatS. PaesanoI. PedruzziS. PilboutL. PourcelS. PouxM. PozzatoManuela PruessNicole RedaschiC. RivoireChristian J A SigristKarin SonessonS. SundaramAnastasia SveshnikovaCathy H. WuC. ArighiChuming ChenYongxing ChenHongzhan HuangK. LaihoMinna LehvaslaihoPeter B. McGarveyD. NataleKaren RossC. R. VinayakaYuqi WangJian Zhang
Nucleic Acids Research Published 2024/11/18
Quantum error correction below the surface code threshold
R. Acharya;Laleh Aghababaie-BeniI. AleinerTrond I. AndersenM. AnsmannF. AruteK. AryaA. AsfawN. AstrakhantsevJ. AtalayaR. BabbushDave BaconB. BallardJ. C. BardinJohannes BauschA. BengtssonA. BilmesS. BlackwellS. BoixoG. BortoliA. BourassaJ. BovairdL. BrillM. BroughtonD. A. BrowneB. BucheaB. BuckleyD. BuellT. BurgerB. BurkettN. BushnellA. CabreraJ. CamperoHung-Shen ChangYu ChenZijun ChenB. ChiaroDesmond ChikCharina ChouJ. ClaesA. ClelandJ. CoganR. CollinsP. ConnerW. CourtneyA. CrookB. CurtinSayan DasA. DaviesL. LorenzoD. DebroyS. DemuraM. DevoretA. PaoloP. DonohoeI. DrozdovA. DunsworthC. EarleT. EdlichA. EickbuschA. M. ElbagM. ElzoukaC. EricksonL. FaoroE. FarhiV. S. FerreiraL. BurgosE. ForatiA. FowlerB. FoxenS. GanjamG. GarcíaR. Gasca'Elie GenoisW. GiangC. GidneyD. GilboaR. GosulaA. DauD. GraumannA. GreeneJ. GrossS. HabeggerJohn HallMichael C. HamiltonM. HansenM. HarriganS. D. HarringtonF. HerasS. HeslinP. HeuO. HiggottG. HillJ. HiltonGeorge HollandSabrina HongHsin-Yuan HuangA. HuffW. HugginsL. IoffeS. IsakovJ. IvelandE. JeffreyZhang JiangCody JonesS. JordanC. JoshiP. JuhásD. KafriHui KangA. KaramlouK. KechedzhiJ. KellyT. KhaireT. KhattarM. KhezriSeon KimP. KlimovA. KlotsB. KobrinPushmeet KohliA. KorotkovF. KostritsaRobin KothariBorislav M. KozlovskiiJ. KreikebaumV. D. KurilovichN. LacroixD. LandhuisT. Lange-DeiB. W. LangleyP. LaptevK. LauL. GuevelJ. LedfordKenny LeeY. LenskyShannon LeonB. LesterWing Yan LiYin LiA. LillWayne LiuW. LivingstonA. LocharlaE. LuceroD. LundahlA. LuntS. MadhukF. MaloneA. MaloneySalvatore Mandr'aL. S. MartinSteven MartinO. MartinC. MaxfieldJ. McCleanM. McEwenS. MeeksA. MegrantX. MiK. MiaoA. MieszalaR. MolaviS. MolinaS. MontazeriA. MorvanR. MovassaghW. MruczkiewiczO. NaamanMatthew NeeleyC. NeillA. NersisyanH. NevenMichael NewmanJ. NgA. NguyenM. NguyenChia-Hung NiT. O’BrienW. D. OliverA. OpremcakK. OttossonA. PetukhovA. PizzutoJohn C. PlattR. PotterO. PritchardL. PryadkoC. QuintanaG. RamachandranM. ReagorD. M. RhodesG. RobertsEliot RosenbergE. RosenfeldP. RoushanN. RubinN. SaeiD. SankK. SankaragomathiK. SatzingerH. SchurkusC. SchusterA. W. SeniorM. ShearnA. ShorterN. ShuttyV. ShvartsShraddha SinghV. SivakJ. SkruznyS. SmallV. SmelyanskiyW. C. SmithR. SommaS. SpringerG. SterlingD. StrainJ. SuchardA. SzaszA. SzteinD. ThorA. TorresM. M. TorunbalciA. VaishnavJ. VargasS. VdovichevG. VidalB. VillalongaC. HeidweillerS. WaltmanShannon WangB. WareKate WeberT. WhiteK. WongB. WooC. XingZ. YaoP. YehB. YingJuhwan YooN. YosriG. YoungAdam ZalcmanYaxing ZhangN. ZhuN. Zobrist
Nature Published 2024/08/24

Summary:

Two below-threshold surface code memories on Willow, a distance-7 code and a distance-5 code integrated with a real-time decoder, indicate device performance that, if scaled, could realize the operational requirements of large-scale fault-tolerant quantum algorithms.

Book review: Christoph Molnar. 2020. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable
R. K. Sinha
Metamorphosis Published 2024/06/01
TRIPOD+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods
Gary S. Collins;K. MoonsP. DhimanR. RileyA. L. BeamBen Van CalsterMarzyeh GhassemiXiaoxuan LiuJohannes B ReitsmaM. van SmedenA. BoulesteixJ. CamaradouL. CeliS. DenaxasA. DennistonBen GlockerRobert M GolubHugh HarveyG. HeinzeMichael M. HoffmanA. KengneEmily LamNaomi LeeElizabeth W LoderLena Maier-HeinB. MateenMelissa MccraddenLauren Oakden-RaynerJohan OrdishRichard ParnellS. RoseKarandeep SinghL. WynantsP. Logullo
The BMJ Published 2024/04/16

Summary:

The development of TRIPOD+AI is described and the expanded 27 item checklist with more detailed explanation of each reporting recommendation is presented, and the TRIPOD+AI for Abstracts checklist is presented.

StarCoder 2 and The Stack v2: The Next Generation
Anton Lozhkov;Raymond LiLoubna Ben AllalFederico CassanoJ. Lamy-PoirierNouamane TaziAo TangDmytro PykhtarJiawei LiuYuxiang WeiTianyang LiuMax TianDenis KocetkovArthur ZuckerYounes BelkadaZijian WangQian LiuDmitry AbulkhanovIndraneil PaulZhuang LiWen-Ding LiMegan L. RisdalJia LiJian ZhuTerry Yue ZhuoEvgenii ZheltonozhskiiNii Osae Osae DadeW. YuLucas KraussNaman JainYixuan SuXuanli HeManan DeyE. AbatiYekun ChaiNiklas MuennighoffXiangru TangMuhtasham OblokulovC. AkikiMarc MaroneChenghao MouMayank MishraA. GuBinyuan HuiTri DaoA. ZebazeOlivier DehaeneN. PatryCanwen XuJulian J. McAuleyHan HuTorsten ScholakSébastien PaquetJennifer RobinsonC. AndersonNicolas ChapadosM. PatwaryNima TajbakhshYacine JerniteCarlos Muñoz FerrandisLingming ZhangSean HughesThomas WolfArjun GuhaL. V. WerraH. D. Vries
ArXiv Published 2024/02/29

Summary:

The BigCode project, an open-scientific collaboration focused on the responsible development of Large Language Models for Code (Code LLMs), introduces StarCoder2, a large model that significantly outperforms other models of comparable size and makes the model weights available under an OpenRAIL license.

OLMo: Accelerating the Science of Language Models
Dirk Groeneveld;Iz BeltagyPete WalshAkshita BhagiaRodney KinneyOyvind TafjordA. JhaHamish IvisonIan MagnussonYizhong WangShane AroraDavid AtkinsonRussell AuthurKhyathi Raghavi ChanduArman CohanJennifer DumasYanai ElazarYuling GuJack HesselTushar KhotWilliam MerrillJacob Daniel MorrisonNiklas MuennighoffAakanksha NaikCrystal NamMatthew E. PetersValentina PyatkinAbhilasha RavichanderDustin SchwenkSaurabh ShahW. SmithEmma StrubellNishant SubramaniMitchell WortsmanPradeep DasigiNathan LambertKyle RichardsonLuke S. ZettlemoyerJesse DodgeKyle LoLuca SoldainiNoah A. SmithHanna Hajishirzi
Published 2024/02/01

Summary:

OLMo is built, a competitive, truly Open Language Model, to enable the scientific study of language models and it is hoped this release will empower the open research community and inspire a new wave of innovation.

TÜLU 3: Pushing Frontiers in Open Language Model Post-Training
Nathan Lambert;Jacob Daniel MorrisonValentina PyatkinShengyi HuangHamish IvisonFaeze BrahmanLester James Validad MirandaAlisa LiuNouha DziriXinxi LyuYuling GuSaumya MalikVictoria GrafJena D. HwangJiangjiang YangRonan Le BrasOyvind TafjordChristopher WilhelmLuca SoldainiNoah A. SmithYizhong WangPradeep DasigiHanna Hajishirzi
ArXiv Published 2024/11/22

Summary:

This work introduces Tulu 3, a family of fully-open state-of-the-art post-trained models, alongside its data, code, and training recipes, serving as a comprehensive guide for modern post-training techniques.

PaliGemma: A versatile 3B VLM for transfer
L. Beyer;A. SteinerAndré Susano PintoAlexander KolesnikovXiao WangDaniel M. SalzMaxim NeumannIbrahim M. AlabdulmohsinMichael TschannenEmanuele BugliarelloThomas UnterthinerDaniel KeysersSkanda KoppulaFangyu LiuAdam GrycnerA. GritsenkoN. HoulsbyManoj KumarKeran RongJulian Martin EisenschlosRishabh KabraMatthias BauerMatko BovsnjakXi ChenM. MindererP. VoigtlaenderIoana BicaIvana BalazevicJ. PuigcerverPinelopi PapalampidiO. HénaffXi XiongRadu SoricutJeremiah HarmsenXiao-Qi Zhai
ArXiv Published 2024/07/10

Summary:

PaliGemma is an open Vision-Language Model that is based on the SigLIP-So400m vision encoder and the Gemma-2B language model that achieves strong performance on a wide variety of open-world tasks.

Continuous 3D Perception Model with Persistent State
Qianqian Wang;Yifei ZhangAleksander HolynskiAlexei A. EfrosAngjoo Kanazawa
2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Published 2025/01/21

Summary:

The model, called CUT3R (Continuous Updating Transformer for 3D Reconstruction), captures rich priors of real-world scenes: not only can it predict accurate pointmaps from image observations, but it can also infer unseen regions of the scene by probing at virtual, unobserved views.

A DNA barcoding framework for taxonomic verification in the Darwin Tree of Life Project
A. Twyford;Jordan BeasleyIan BarnesHeather AllenFreja AzzopardiD. BellM. BlaxterG. BroadL. Campos-DomínguezDarren ChooneaLiam M. CrowleyPiotr CuberMichael CunliffeAlexandra DombrowskiBrian DouglasL. ForrestEster GayaClementine GreevesClaire GriffinJoanna HarleyMichelle L HartPeter W. H. HollandPeter M. HollingsworthInez JanuszczakAmanda L JonesPaul J. KerseyEstelle KiliasM. LawniczakOwen T. LewisSahr MianAlice MinottoRaju MisraPeter O. MulhairLyndall Pereira da ConceicoaBenjamin W. PriceSilvia SalatinoFelix ShawOlga SivellLaura SivessRebekka UhlKieran Woof
Wellcome Open Research Published 2024/06/24

Summary:

The development and implementation of DNA barcoding for the Darwin Tree of Life Project (DToL), which aims to sequence and assemble high quality reference genomes for all eukaryotic species in Britain and Ireland, is described.

ARTIFICIAL INTELLIGENCE IN EDUCATION: CHALLENGES AND OPPORTUNITIES FOR SUSTAINABLE DEVELOPMENT
S. A. Vakhabova;Valery V. KosulinAna Zizaeva
EKONOMIKA I UPRAVLENIE: PROBLEMY, RESHENIYA

Summary:

The results show the significant potential of AI in personalizing learning, automating routine tasks, and providing access to knowledge, but also reveal serious risks of exacerbating social inequality and ethical dilemmas.

The Galaxy platform for accessible, reproducible, and collaborative data analyses: 2024 update
Linelle Ann L Enis Olivier Ahmed H Wendi A Dannon Madeline Abueg Afgan Allart Awan Bacon Baker Bassetti Batut;Linelle AbuegE. AfganOlivier AllartA. H. AwanW. BaconD. BakerMadeline E. BassettiBérénice BatutMatthias BerntDaniel J. BlankenbergA. BombarelyAnthony BretaudeauCatherine J. BromheadMelissa L BurkePatrick K. CaponMartin ČechMaría Chavero-DíezJ. ChiltonTyler CollinsFrederik CoppensNate CoraorG. CuccuruFabio CumboJohn DavisPaul F De GeestW. de KoningM. DemkoAssunta D. DesantoJ. D. BeginesMaria A. DoyleBert DroesbekeAnika Erxleben-EggenhoferM. FöllGiulio FormentiA. FouillouxRendani GangazheTanguy GenthonJeremy GoecksA. G. BeltránN. GoonasekeraNadia GouéTim J. GriffinBjörn A. GrüningAysam GuerlerSveinung GundersenOve Johan Ragnar GustafssonChristina R. HallThomas W HarropHelge HechtA. HeidariTillman HeisnerF. HeylSaskia D. HiltemannH. HotzCameron J HydeP. JagtapJulia JakielaJames E. JohnsonJayadev JoshiMarie JosséKhaled Jum’ahMatúš KalašKatarzyna KamienieckaTunc KayikciogluM. KonkolLeonid KostrykinNatalie KucherAnup KumarM. KuntzDelphine LarivièreRoss LazarusY. L. BrasGildas Le CorguilléJustin LeeSimone LeoLeandro LiborioRomane LiboubanD. TaberneroLucille Lopez-DelisleLaila S LosAlexandru MahmoudI. MakuninP. MarinSubina P. MehtaWinnie MokPablo MorenoFrançois Morier-GenoudStephen MosherTeresa MüllerEngy NasrA. NekrutenkoTiffanie M. NelsonAsime ObaAlexander E. OstrovskyPolina V PoluninaKrzysztof PoterlowiczE. PriceGareth R PriceH. RascheBryan RaubenoltColine RoyauxLuke SargentMichelle T SavageVolodymyr S. SavchenkoD. SavchenkoMichael C. SchatzPauline SeguineauBeatriz Serrano-SolanoNicola SoranzoSanjay Kumar SrikakulamKeith SudermanAnna SymeM. TangaroJonathan TeddsM. TekmanWai Cheng (Mike) ThangAnil S. ThankiM. UhlMarius van den BeekDeepti VarshneyJennifer VessioPavankumar VidemGreg Von KusterG. R. WatsonNatalie Whitaker-AllenUwe WinterM. WolstencroftF. ZambelliP. ZierepRand Zoabi
Nucleic Acids Research Published 2024/05/20

Summary:

Code development continues in line with the Galaxy Project roadmap, with improvements to job scheduling and the user interface, and general purpose graphical processing units (GPGPU) access for cutting-edge methods, and licensed tool support.

InterPro: the protein sequence classification resource in 2025
Matthias Blum;Antonina AndreevaLaise Cavalcanti FlorentinoS. ChuguranskyTiago GregoEmma E. M. HobbsBeatriz Lázaro PintoAilsa OrrT. Paysan-LafosseIrina PonamarevaGustavo A. SalazarNicola BordinP. BorkA. BridgeLucy J ColwellJulian GoughD. HaftIvica LetunicFelipe Llinares-LópezAron Marchler-BauerLaetitia Meng-PapaxanthosH. MiD. NataleChristine A OrengoA. P. PanduranganD. PiovesanC. RivoireChristian J. A. SigristN. ThankiF. Thibaud-NissenP. D. ThomasSilvio C. E. TosattoCathy H. WuA. Bateman
Nucleic Acids Research Published 2024/11/20

Summary:

The status of InterPro is reported on, detailing new developments in the database, associated web interface and software, including the increased integration of structures predicted by AlphaFold and the enhanced description of protein families using artificial intelligence.

AlphaEvolve: A coding agent for scientific and algorithmic discovery
Alexander Novikov;Ngân V˜uMarvin EisenbergerEmilien DupontPo-Sen HuangAdam Zsolt WagnerS. ShirobokovBorislav M. KozlovskiiFrancisco J. R. RuizAbbas MehrabianM. P. KumarAbigail SeeSwarat ChaudhuriGeorge HollandA. DaviesSebastian NowozinPushmeet KohliMatej BalogGoogle DeepMind
ArXiv Published 2025/06/16

Summary:

AlphaEvolve is an evolutionary coding agent that substantially enhances capabilities of state-of-the-art LLMs on highly challenging tasks such as tackling open scientific problems or optimizing critical pieces of computational infrastructure.

Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs
Abdelrahman Abouelenin;Atabak AshfaqAdam AtkinsonH. AwadallaNguyen BachJianmin BaoAlon BenhaimMartin CaiVishrav ChaudharyCongcong ChenDongdong ChenDongdong ChenJunkun ChenWeizhu ChenYen-Chun ChenYi-Ling ChenQi DaiXiyang DaiRuchao FanMei GaoMingcheng GaoAmit GargAbhishek GoswamiJunheng HaoAmr HendyYuxuan HuXin JinMahmoud KhademiDongwoo KimYoung Jin KimGina LeeJinyu LiYunsheng LiChen LiangXihui LinZeqi LinMeng-Jie LiuYang LiuGilsinia LopezChong LuoPiyush MadanV. MazalovAli MousaviA. NguyenJing PanD. Perez-BeckerJacob PlatinThomas PortetKai QiuBo RenLiliang RenSambuddha RoyNing ShangYelong ShenSaksham SinghalSubhojit SomXiaocheng SongT. SychPraneetha VaddamanuShuohang WangYiming WangZhenghao WangHaibin WuHaoran XuWeijian XuYifan YangZiyi YangDonghan YuI. ZabirJianwen ZhangL. ZhangYunan ZhangXiren Zhou
ArXiv Published 2025/03/03

Summary:

Despite its compact 3.8-billion-parameter size, this experimental version of Phi-4-Mini achieves reasoning performance on par with or surpassing significantly larger models, including DeepSeek-R1-Distill-Qwen-7B and DeepSeek-R1-Distill-Llama-8B.

A benchmark of expert-level academic questions to assess AI capabilities
Long Phan;Alice GattiZiwen HanNathaniel LiJosephina HuHugh ZhangChen Bo Calvin ZhangSean ShiMichael ChoiAnish AgrawalArnav ChopraAdam KhojaRichard RenRyan KimJ. HausenloyOliver ZhangMantas MazeikaDaron AndersonTung NguyenImad Ali ShahMobeen MahmoodFiona FengSteven Y. FengHaoran ZhaoMichael YuVarun Prashant GangalChelsea ZouZihan WangJessica WangPawan KumarOleksandr PokutnyiRobert GerbiczS. PopovJohn-Clark LevinMstyslav KazakovJohannes SchmittGeoff GalgonÁ. SánchezYongki LeeWill YeadonScott C. SauersMarc RothC. AguSøren RiisFabian GiskaSaiteja UtpalaZachary GiboneyG. M. GoshuJ. XavierSarah-Jane CrowsonM. NaiyaNoah BurnsLennart FinkeZerui ChengHyunwoo ParkFrancesco Fournier-FacioJohn B. WydallisM. NandorAnkit SinghTim GehrungerJiaqi CaiBen McCartyD. DucloselJungbae NamJennifer A. ZampeseRyan G. HoerrAras BachoGautier Abou LoumeAbdallah GalalHangrui CaoAlexis C GarretsonDamien SileoQiuyu RenDoru CojocPavel ArkhipovUsman QaziLianghui LiS. MotwaniChristian Schröder de WittEdwin TaylorJohannes VeithEric SingerTaylor D. HartmanP. RissoneJaehyeok JinJ. ShiChris G. WillcocksJoshua RobinsonAleksandar MikovAmeya PrabhuLongke TangXavier AlapontJustine Leon UroKevin ZhouEmily de Oliveira SantosAndrey Pupasov MaksimovEdward VendrowKengo ZenitaniJulien GuillodYuqi LiJoshua VendrowVladyslav M. KuchkinZe-An NgPierre MarionDenis EfremovJayson LynchKaiqu LiangAndrew GritsevskiyDakotah MartinezB. PagelerNick CrispinoD. ZvonkineNatanael Wildner FragaSaeed SooriO. PressHenry TangJulian SalazarSean R. GreenLina BrusselMoon TwayanaAymeric DieuleveutT. RogersWenjin ZhangBikun LiJinzhou YangArun RaoGabriel LoiseauMikhail KalininMarco LukasCiprian ManolescuSubrata MishraAriel Ghislain Kemogne KamdoumT. KreimanTad HoggAlvin JinCarlo BosioGongbo SunBrian P. CoppolaTim TarverH. HeidingerRafael SayousStefan IvanovJoe CavanaghJiawei ShenJoseph Marvin ImperialPhilippe SchwallerShaipranesh SenthilkumaAndrés M BranA. DehghanA. AlgabaBrecht VerbekenDavid A. NoeverV. RagavendranPLisa SchutIlia SucholutskyEvgenii ZheltonozhskiiDerek LimRichárd StanleyShankar SivarajanTong YangJohn MaarJulian WykowskiMartí OllerJennifer SandlinAnmol SahuYuzheng HuSara FishN. HeydariArchimedes T. AprontiKai RawalTobías García VilchisYuexuan ZuM. LacknerJames KoppelJeremy NguyenD. AntonenkoSteffi ChernBingchen ZhaoPierrot ArseneA. GoldfarbSergey IvanovRafal PoswiataChenguang WangDaofeng LiDonato CrisostomiDonato CrisostomiBenjamin MyklebustArchan SenD. PerrellaNurdin KaparovMark H InlowAllen ZangElliott ThornleyDaniil OrelVladislav PoritskiS. Ben-DavidZachary BergerParker WhitfillMichael FosterD. MunroLinh HoDan Bar HavaAleksey KuchkinRobert LauffDavid HolmesF. SommerhageK. SchneiderZ. KazibweNate StambaughMukhwinder SinghI. MagoulasD. ClarkeD. KimF. DiasDae Hyun KimK. AgarwalVictor Efren Guadarrama VilchisImmo KloseChristoph DemianUjjwala AnantheswaranAdam ZweigerGuglielmo AlbaniJeffery LiNicolas DaansM. RadionovVáclav RozhoňZiqiao MaChristian StumpMohammed BerkaniJacob PlatnickVolodymyr NevirkovetsLuke BaslerM. PiccardoF. JeanplongNiv CohenJosef TkadlecP. RosuPiotr PadlewskiStanislaw BarzowskiKyle MontgomeryAline MenezesArkil PatelZixuan WangJamie Tucker-FoltzJack StadeTom GoertzenF. KazemiJeremiah MilbauerJohn Arnold AmbayJeremiah MilbauerYan Carlos Leyva LabradorAlan Givr'eHew WolffV. RossbachM. AzizYounesse KaddarYanxu ChenRobin ZhangJiayi PanAntonio TerpinJiayi PanHailey SchoelkopfE. ZhengAvishy CarmiAdam JonesJainam ShahEthan D. L. BrownKelin ZhuM. BartoloRichard S. WheelerAndrew HoShaul BarkanJiaqi WangMartin StehbergerEgor KretovKaustubh SridharKaustubh SridharAnji ZhangD. PydaJoanna TamDavid M. CunninghamV. GoryachevDemosthenes PatramanisMichael KrauseAndrew RedentiDaniel BugasDavid AldousJesyin LaiShannon ColemanMohsen BahalooJiangnan XuSangwon LeeSandy ZhaoNing TangMichael K. CohenMicah CarrollMicah CarrollJ. KirchnerStefan SteinerbergerM. OvchynnikovJason O. MatosAdithya ShenoyBenedito Alves de Oliveira JuniorMichael WangYuzhou NiePaolo GiordanoPhilipp PetersenAnna Sztyber-BetleyPriti ShuklaJ. CrozierAntonella PintoShreyas VermaPrashant JoshiZheng-Xin YongAllison TeeZheng-Xin YongOrion WellerRaghav SinghalOrion WellerA. IvanovSeri KhouryH. MostaghimiKunvar ThamanQijia ChenTran Quoc Kh'anhJacob LoaderStefano CavalleriHannah SzlykZachary BrownJonathan RobertsWilliam AlleyKunyang SunRyan StendallMax LamparthAnka ReuelTing WangHanmeng XuSreenivas Goud RaparthiPablo Hern'andez-C'amaraFreddie MartinDmitry MalishevT. PreuTomasz KorbakMarcus AbramovitchDominic J. WilliamsonZiye ChenBir'o B'alintM Saiful BariP. KassaniZihao WangB. AnsarinejadLaxman Prasad GoswamiYewen SunHossam ElgnainyD. TorderaGeorge BalabanianEarth AndersonLynna KvistadAlejandro Jos'e MoyanoRajat MaheshwariAhmad SakorMurat EronI. McAlisterJavier GimenezInnocent EnyekweAndrew Favre D.O.Shailesh ShahXiaoxiang ZhouFiruz KamalovRonald ClarkSherwin AbdoliTim SantensKhalida MeerHarrison K WangK. RamakrishnanEvan ChenA. TomasielloG. LucaShi-Zhuo LooiVinh-Kha LeNoam KoltNiels MundlerAvi SemlerNiels MündlerJacob DroriCarl J FossumMilind JagotaRonak PradeepHonglu FanTej ShahJonathan EicherMichael ChenKushal ThamanWilliam MerrillCarter HarrisJason GrossIlya GusevAsankhaya SharmaShashank AgnihotriP. ZhelnovSiranut UsawasutsakornMohammadreza MofayeziSergei BogdanovA. PiperskiMarc CarauleanuDavid K. ZhangDylan LerRoman LeventovIgnat SorokoThorben JansenPascal LauerJoshua DuerschVage TaamazyanWiktor MorakWenjie MaWilliam HeldT. HuyRuicheng XianA. ZebazeMohanad MohamedJulian Noah LeserMichelle YuanLaila YacarJohannes LenglerHossein ShahrtashEdson OliveiraJoseph W. JacksonD. GonzalezAndy ZouMuthu ChidambaramTimothy ManikHector HaffendenDashiell StanderA. DasouqiAlexander ShenEmilien DucBita GolshaniDavid StapMikalai UzhouAlina B. ZhidkovskayaLukas LewarkM'aty'as VinczeDustin WehrColin TangZaki HossainShaun PhillipsMuzhen JiangFredrik EkstromAngela HammonOam PatelNicolas RemyFaraz FarhidiG. MedleyF. MohammadzadehMadellene PeñaflorHaile KassahunAlena FriedrichClaire SparrowTaom SakalOmkar DhamaneAli Khajegili MirabadiE. HallmanM. BattagliaMohammad MaghsoudimehrabaniHieu HoangAlon AmitDave HulbertR. PereiraS. WeberStephen MensahNathan AndreAnton PeristyyChris HarjadiHimanshu GuptaS. MalinaHimanshu GuptaWill CaiMustafa MehkaryFrank ReidegeldAnna-Katharina DickCary FridayJasdeep SidhuWanyoung KimM. CostaHubeyb GurdoganBrian WeberHarsh KumarTong JiangArunim AgarwalChiara CeconelloW. S. VazChao ZhuangHaon ParkA. TawfeekHaon ParkMichael KirchhofLinjie DaiMichael KirchhofJohan FerretYuzhou Wang
Nature Published 2025/01/24
DeepSeek-V3.2: Pushing the Frontier of Open Large Language Models
DeepSeek-AI;A. LiuAoxue MeiB. LinBing XueBing-Li WangBin XuBochao WuBowei ZhangChaofan LinChen DongChengda LuChenggang ZhaoC. DengChenhao XuC. RuanDamai DaiDaya GuoDejian YangDeli ChenErhang LiFangqi ZhouFangyun LinFucong DaiGuangbo HaoGuanting ChenGuowei LiH. ZhangHanwei XuHao LiHao LiangHaoran WeiHaowei ZhangHao-sheng LuoHaozhe JiHonghui DingHongxuan TangHuan CaoHuazuo GaoHuixian QuHui ZengJialiang HuangJiashi LiJiaxin XuJiewen HuJingChang ChenJi XiangJingyang YuanJing ChengJinhua ZhuJun RanJunguang JiangJunjie QiuJunlong LiJun-Mei SongKai DongKaige GaoKang GuanKexin HuangKexin ZhouKe-wei HuangK. YuLean WangLecong ZhangLei WangLiang ZhaoLiangsheng YinLihua GuoLing-Li LuoLin MaLitong WangLiyue ZhangM. DiM. Y. XuMingchuan ZhangMinghua ZhangM. TangMingxu ZhouP. HuangPeixin CongPeiyi WangQiancheng WangQihao ZhuQingyang LiQinyu ChenQiushi DuRuiling XuRuiqi GeRuisong ZhangRuizhe PanRunji WangRunqiu YinRunxin XuRu ShenRuoyu ZhangS. LiuShanghao LuShangyan ZhouShanhuang ChenShaofei CaiShaoyuan ChenShengding HuShengyu LiuShiqiang HuShirong MaShiyu WangShuiping YuShunfeng ZhouShuting PanSongyang ZhouTao NiTao YunTian PeiTian YeTianyuan YueWangding ZengWen LiuW. LiangWenjie PangWenjing LuoWenjun GaoWentao ZhangXi GaoXiangwen WangXiaoling BiXiaodong LiuXiaohan WangXiaokang ChenXiaokang ZhangX. NieXin ChengXin LiuXin XieXingchao LiuXingkai YuXingyou LiXinyu YangXinyuan LiXu ChenXuecheng SuXuehai PanXuheng LinXu FuY. Q. WangYang ZhangYanhong XuYanru MaYao LiYao ZhaoYaofeng SunYaohui WangY. QianYingpu YuYichao ZhangYifan DingYifan ShiYi XiongYing HeYing ZhouYinmin ZhongY. PiaoYisong WangYixiao ChenYixuan TanYixuan WeiYiyang MaYiyuan LiuYong YangYongqiang GuoYongtong WuYu WuYuan ChengY. OuYuanfang XuYuduan WangYue GongYuhan WuYu-Hui ZouYukun LiYunfan XiongYu-Wei LuoYu-mei YouYuxuan LiuYuyang ZhouZ. F. WuZ. RenZehua ZhaoZ. RenZhangli ShaZhe FuZhean XuZhenda XieZhen-guo ZhangZhewen HaoZhibin GouZhicheng MaZhigang YanZhihong ShaoZhixian HuangZhiyu WuZhuoshu LiZhuping ZhangZian XuZihao WangZihui GuZijia ZhuZi-Rui LiZipeng ZhangZiwei XieZiyi GaoZizheng PanZ. YaoB. FengHui LiJ. CaiJ. NiLei XuMeng LiNing TianR. J. ChenR. JinS. LiShuang ZhouT. SunX. Q. LiXiangyu JinXiaojin ShenXiaosha ChenXinnan SongXinyi ZhouY. X. ZhuYanping HuangYao LiYi ZhengYuchen ZhuYunxiang MaZhen HuangZhipeng XuZhongyu ZhangDong-Li JiJian LiangJianzhong GuoJin ChenLeyi XiaMiaojun WangMingming LiPeng ZhangRuyi ChenShangmian SunShao-Kang WuSheng-Ying YeT.WangW. XiaoWei AnXianzu WangXiaowen SunXiaoxiang WangYing TangY. ZhaZe-Na ZhangZhenghua JuZhen ZhangZ. Qu
ArXiv Published 2025/12/02

Summary:

This work introduces DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance, and introduces DSA, an efficient attention mechanism that substantially reduces computational complexity while preserving model performance in long-context scenarios.

Why Do Multi-Agent LLM Systems Fail?
M. Cemri;Melissa Z. PanShuyi YangLakshya A. AgrawalBhavya ChopraRishabh TiwariKurt KeutzerAditya G. ParameswaranDan KleinK. RamchandranMatei ZahariaJoseph E. GonzalezIon Stoica
ArXiv Published 2025/03/17

Summary:

This work builds the first Multi-Agent System Failure Taxonomy (MAST), a comprehensive dataset of 1600+ annotated traces collected across 7 popular MAS frameworks, and develops an LLM-as-a-Judge pipeline with high agreement with human annotations to enable scalable annotation.

The T2K experiment
T. Abe;N. AbgrallH. AiharaY. AjimaJ. AlbertD. AllanP. AmaudruzC. AndreopoulosB. AndrieuM. AnerellaC. AngelsenS. AokiO. AraokaJ. ArgyriadesA. ArigaT. ArigaS. AssylbekovJ. Andr'eD. AutieroA. BadertscherO. BallesterM. BarbiG. BarkerP. BaronG. BarrL. BartoszekM. BatkiewiczF. BayS. BenthamV. BerardiB. BergerH. BernsI. BertramM. BesnierJ. BeucherD. BeznoskoS. BhadraP. BirneyD. BishopE. BlackmoreF. BlaszczykJ. BłockiA. BlondelA. BodekC. BojechkoJ. BouchezT. BoussugeS. BoydM. BoyerN. BraamR. BradfordA. BravarK. BriggsJ. BrinsonC. BronnerD. Brook-RobergeM. BryantN. BuchananH. BuddM. CadabeschiR. CallandD. CalvetJ. Rodr'iguezJ. CarrollS. CartwrightA. CarverR. CastilloM. CatanesiC. CavataA. CazesA. CerveraJ. CharrierC. ChávezS. ChoiS. CholletG. ChristodoulouP. ColasJ. ColemanW. ColemanG. CollazuolK. ConnollyP. CookeA. CurioniA. DabrowskaI. DankóR. DasG. DaviesS. DavisM. DayX. BroiseP. PerioG. RosaT. DealtryA. DebraineE. DelagnesA. DelbartC. DenshamF. LodovicoS. LuiseP. TranJ. DobsonJ. DoornbosU. DoreO. DrapierF. DruilloleF. DufourJ. DumarchezT. DurkinS. DytmanM. DziewieckiM. DziombaB. EllisonS. EmeryA. EreditatoJ. EscallierL. EscuderoL. EspositoW. FaszerM. FechnerA. FerreroA. FinchC. FisherM. FittonR. FlightD. ForbushE. FrankK. FranshamY. FujiiY. FukudaM. GallopV. GalymovG. GanetisF. GannawayA. GaudínJ. GawedaA. GendottiM. GeorgeS. GiffinC. GigantiK. GiljeI. GiomatarisJ. GiraudA. GhoshT. GolanM. GoldhaberJ. Gomez-CadenasS. GomiM. GoninM. GoyetteA. GrantN. GrantF. GrañenaS. GreenwoodP. GumplingerP. GuzowskiM. HaighK. HamanoC. HansenT. HaraP. HarrisonB. HartfielM. HartzT. HaruyamaR. HasanenT. HasegawaN. HastingsS. HastingsA. HatzikoutelisK. HayashiY. HayatoT. HaycockC. HeartyR. HelmerR. HendersonS. HerlantN. HigashiJ. HignightK. HiraideE. HiroseJ. HoleczekN. HonkanenS. HorikawaA. HyndmanA. IchikawaK. IekiM. IevaM. IidaM. IkedaJ. IlicJ. ImberT. IshidaC. IshiharaT. IshiiS. IvesM. IwasakiK. IyogiA. IzmaylovB. JamiesonR. JohnsonK. JooG. Jover-ManasC. JungH. KajiT. KajitaH. KakunoJ. KamedaK. KaneyukiD. KarlenK. KasamiV. KaseyI. KatoH. KawamukoE. KearnsL. KelletM. KhabibullinM. KhaleeqN. KhanA. KhotjantsevD. KiełczewskaT. KikawaJ. KimS. KimN. KimuraB. KirbyJ. KisielP. KitchingT. KobayashiG. KoganS. KoikeT. KomorowskiA. KonakaL. KormosA. KorzenevK. KosekiY. KoshioY. KouzumaK. KowalikV. KravtsovI. KresloW. KroppH. KuboJ. KubotaY. KudenkoN. KulkarniL. KurchaninovY. KurimotoR. KurjataY. KurosawaT. KutterJ. LagodaK. LaihemR. LangstaffM. LavederT. LawsonP. T. LeA. CoguieM. L. RossK. P. LeeM. LenckowskiC. LicciardiI. LimT. LindnerR. P. LitchfieldA. LonghinG. LópezP. LuL. LudoviciT. LuxM. MacaireL. MagalettiK. MahnY. MakidaC. J. MalafisM. MałekS. ManlyA. MarchionniC. MarkA. MarinoA. MaroneJ. MarteauJ. MartinT. MaruyamaT. MaryonJ. MarzecP. MasliahE. MathieC. MatsumuraK. MatsuokaV. MatveevK. MavrokoridisE. MazzucatoN. McCauleyK. McFarlandC. McgrewT. McLachlanI. MercerM. MessinaW. MetcalfC. MetelkoM. MezzettoP. MijakowskiC. A. MillerA. MinaminoO. MineevS. MineR. MinvielleG. MitukaM. MiuraK. MizouchiJ. MolsL. MonfregolaE. MonmartheF. MoreauB. MorganS. MoriyamaD. MorrisA. MuirA. MurakamiJ. MuratoreM. MurdochS. MurphyJ. MyslikG. NagashimaT. NakadairaM. NakahataT. NakamotoK. NakamuraS. NakayamaT. NakayaD. NaplesB. NelsonT. NichollsK. NishikawaH. NishinoK. NittaF. NizeryJ. NowakM. NoyY. ObayashiT. OgitsuH. OhhataT. OkamuraK. OkumuraT. OkusawaC. OhlmannK. OlchanskiR. OpenshawS. OserM. OtaniR. OwenY. OyamaT. OzakiM. PacV. PalladinoV. PaoloneP. PaulD. PayneG. PearceC. PearsonJ. PerkinM. PflegerF. PierreD. PierrepontP. PlonskiP. PoffenbergerE. PopławskaB. PopovM. PosiadałaJ. PoutissouR. PoutissouR. PreeceP. PrzewłockiW. QianJ. RaafE. RadicioniK. RamosP. RatoffT. RauferM. RavonelM. RaymondF. RetièreD. RichardsJ. RitouA. RobertP. RodriguesE. RondioM. RoneyM. RooneyD. RossB. RossiS. RothA. RubbiaD. RuterboriesR. SaccoS. SadlerK. SakashitaF. SánchezA. SarratK. SasakiP. SchaackJ. SchmidtK. ScholbergJ. SchwehrM. ScottD. ScullyY. SeiyaT. SekiguchiH. SekiyaG. ShefferM. ShibataY. ShimizuM. ShiozawaS. ShortM. SiyadD. SmithR. SmithM. SmyJ. SobczykH. SobelS. SooriyakumaranM. SorelJ. SpitzA. StahlP. StamoulisO. StarJ. StatterL. StawnyczyJ. SteinmannJ. SteffensB. StillM. StodulskiJ. StoneC. StrabelT. StraussR. SulejP. SutcliffeA. SuzukiK. SuzukiS. SuzukiS. SuzukiY. SuzukiJ. SwierblewskiT. SzegłowskiM. SzeptyckaR. TacikM. TadaA. TadepalliM. TaguchiS. TakahashiA. TakedaY. TakenagaY. TakeuchiH. TanakaK. TanakaM. TanakaM. TanakaN. TanimotoK. TashiroI. TaylorA. TerashimaD. TerhorstR. TerriL. ThompsonA. ThorleyM. ThorpeW. TokiT. TomaruY. TotsukaC. TouramanisT. TsukamotoV. TvaskisM. TzanovY. UchidaK. UenoM. UsseglioA. VacheretM. VaginsJ. SchalkwykJ. VanelG. VasseurO. VeledarP. VincentT. WachalaA. WaldronC. WalterP. WandererM. WardG. WardD. WarkD. WarnerM. WasckoA. WeberR. WendellJ. WendlandN. WestL. WhiteheadG. WikstromR. WilkesM. WilkingZ. Williamson
Scholarpedia
Artificial intelligence, firm growth, and product innovation
T. Babina;A. FedykA. HeJames Hodson
Journal of Financial Economics Published 2024/01/01

Summary:

A new measure of firm-level AI investments is proposed, using a unique combination of worker resume and job postings datasets, which reveals a stark increase in AI investments across sectors.

GPT-4 passes the bar exam
D. Katz;M. BommaritoShang GaoPablo Arredondo
Philosophical transactions. Series A, Mathematical, physical, and engineering sciences Published 2024/02/26

Summary:

GPT-4 significantly outperforms both human test-takers and prior models, demonstrating a 26% increase over ChatGPT and beating humans in five of seven subject areas, document not just the rapid and remarkable advance of large language model performance generally, but also the potential for such models to support the delivery of legal services in society.

A foundation model for the Earth system
Cristian Bodnar;W. BruinsmaAna LucicMegan StanleyAnna AllenJohannes BrandstetterPatrick GarvanMaik RiechertJonathan A. WeynHaiyu DongJayesh K. GuptaKit ThambiratnamAlexander T. ArchibaldChun-Chieh WuE. HeiderMax WellingRichard E. TurnerP. Perdikaris
Nature Published 2025/05/21

Summary:

Aurora, a large-scale foundation model trained on more than one million hours of diverse geophysical data, outperforms operational forecasts in predicting air quality, ocean waves, tropical cyclone tracks and high-resolution weather, all at orders of magnitude lower computational cost.

DeepSeek-VL2: Mixture-of-Experts Vision-Language Models for Advanced Multimodal Understanding
Zhiyu Wu;Xiaokang ChenZizheng PanXingchao LiuWen LiuDamai DaiHuazuo GaoYiyang MaChengyue WuBing-Li WangZhenda XieYu WuKai HuJiawei WangYaofeng SunYukun LiY. PiaoKang GuanA. LiuXin XieYu-mei YouKaihong DongXingkai YuHaowei ZhangLiang ZhaoYisong WangC. Ruan
ArXiv Published 2024/12/13
Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach
Jonas Geiping;Sean McLeishNeel JainJohn KirchenbauerSiddharth SinghBrian R. BartoldsonB. KailkhuraA. BhateleTom Goldstein
ArXiv Published 2025/02/07

Summary:

A novel language model architecture that is capable of scaling test-time computation by implicitly reasoning in latent space by iterating a recurrent block, thereby unrolling to arbitrary depth at test-time is studied.

Kimi-Audio Technical Report
KimiTeam;Ding DingZeqian JuYichong LengSongxiang LiuTong LiuZeyu ShangKai ShenWei SongXu TanHeyi TangZhengtao WangC. WeiYifei XinXinran XuJian-Xiu YuYutao ZhangXinyu ZhouY. CharlesJun ChenYanru ChenYulun DuWeiran HeZhenxing HuGuo-Rong LaiQingchen LiYangyang LiuWeidong SunJianzhou WangYuzhi WangYue WuYuxin WuDongchao YangHaodong YangYingbo YangZhilin YangA. YinRui YuanYutong ZhangZaida Zhou
ArXiv Published 2025/04/25

Summary:

Extensive evaluation shows that Kimi-Audio achieves state-of-the-art performance on a range of audio benchmarks including speech recognition, audio understanding, audio question answering, and speech conversation.

BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions
Terry Yue Zhuo;Minh Chien VuJenny ChimHan HuWenhao YuRatnadira WidyasariImam Nur Bani YusufHaolan ZhanJunda HeIndraneil PaulSimon BrunnerChen GongThong HoangA. ZebazeXiao-ke HongWen-Ding LiJean KaddourMinglian XuZhihan ZhangPrateek YadavNaman JainAlex GuZhoujun ChengJiawei LiuQian LiuZijian WangDavid LoBinyuan HuiNiklas MuennighoffDaniel FriedXiao-Nan DuH. D. VriesL. V. Werra
ArXiv Published 2024/06/22

Summary:

An extensive evaluation of 60 LLMs shows that LLMs are not yet capable of following complex instructions to use function calls precisely, with scores up to 60%, significantly lower than the human performance of 97%, which underscores the need for further advancements in this area.

Towards an AI co-scientist
Juraj Gottweis;Wei-Hung WengAlexander DaryinTao TuAnil PalepuPetar SirkovicArtiom MyaskovskyFelix WeissenbergerKeran RongRyutaro TannoKhaled SaabD. PopoviciJacob BlumFan ZhangKatherine ChouAvinatan HassidimBurak GokturkAmin VahdatPushmeet KohliYossi MatiasA. CarrollKavita KulkarniNenad TomaševYuan GuanVikram DhillonE. D. VaishnavByron LeeTiago R D CostaJos'e R Penad'esGary PeltzYunhan XuAnnalisa PawloskyA. KarthikesalingamVivek Natarajan
ArXiv Published 2025/02/26

Summary:

An AI co-scientist is introduced, a multi-agent system built on Gemini 2.0 intended to help uncover new, original knowledge and to formulate demonstrably novel research hypotheses and proposals, building upon prior evidence and aligned to scientist-provided research objectives and guidance.

The TRIPOD-LLM reporting guideline for studies using large language models
Jack Gallifant;Majid AfsharSaleem AmeenYindalon AphinyanaphongsShan ChenG. CacciamaniDina Demner-FushmanD. DligachRoxana DaneshjouChrystinne FernandesLasse Hyldig HansenAdam LandmanL. LehmannLiam G. McCoyTimothy A. MillerAmy MorenoNikolaj MunchDavid RestrepoG. SavovaR. UmetonJ. GichoyaG. CollinsK. MoonsL. A. CeliD. Bitterman
Nature Medicine Published 2025/01/01

Summary:

Transparent reporting of a multivariable model for individual prognosis or diagnosis–large language model TRIPOD-LLM is a checklist of items considered essential for good reporting of studies that are developing or evaluating an LLM for use in healthcare settings, a ‘living guideline’ that emphasizes transparency, human oversight and task-specific performance reporting.

Simulating 500 million years of evolution with a language model.
Thomas Hayes;Roshan RaoHalil AkinNicholas J SofroniewDeniz OktayZeming LinRobert VerkuilVincent Q TranJonathan DeatonMarius WiggertRohil BadkundriIrhum ShafkatJun GongAlexander DerryRaul S MolinaNeil ThomasYousuf A. KhanChetan MishraCarolyn KimLiam J. BartieMatthew NemethPatrick D. HsuTom SercuSalvatore CandidoAlexander Rives
Science Published 2025/01/16

Summary:

It is shown that language models trained at scale on evolutionary data can generate functional proteins that are far away from known proteins, and ESM3, a frontier multimodal generative language model that reasons over the sequence, structure, and function of proteins is presented.

Deep learning and computer vision in plant disease detection: a comprehensive review of techniques, models, and trends in precision agriculture
Abhishek Upadhyay;N. ChandelKrishna Pratap SinghS. ChakrabortyB. M. NandedeMohit KumarA. SubeeshKonga UpendarAli SalemAhmed Elbeltagi
Artificial Intelligence Review Published 2025/01/17

Summary:

This study reviews the techniques and tools used for automatic disease identification, state-of-the-art DL models, and recent trends in DL-based image analysis, and evaluates various DL architectures, providing guidance on the suitability of these models for production environments.

Who owns this site?

Article Galaxy Pages is a free service from Research Solutions, a company that offers access to content in collaboration with publishing partners, online repositories and discovery services.